# PubMed Pre-training

Biomistral 7B DARE AWQ QGS128 W4 GEMM
An open-source large language model for the medical domain based on the Mistral architecture, further pre-trained on PubMed Central texts
Large Language Model Transformers
B
BioMistral
135
3
Biomistral 7B
Apache-2.0
BioMistral is an open-source large language model optimized for the medical domain based on the Mistral architecture, further pre-trained on PubMed Central open-access text data, supporting multilingual medical question-answering tasks.
Large Language Model Transformers Supports Multiple Languages
B
BioMistral
22.59k
428
Medcpt Article Encoder
Other
MedCPT is a model capable of generating biomedical text embeddings, particularly suitable for semantic search (dense retrieval) tasks.
Text Embedding Transformers
M
ncbi
14.37k
24
Bioelectra PICO
BioELECTRA is a biomedical domain-specific language model pre-trained based on the ELECTRA framework, setting performance records on various biomedical NLP tasks
Large Language Model Transformers
B
kamalkraj
10.88k
8
Ner Disease Ncbi Bionlp Bc5cdr Pubmed
Apache-2.0
Named entity recognition model trained on NCBI Disease dataset and BC5CDR dataset, specialized in identifying disease entities in biomedical literature
Sequence Labeling Transformers Supports Multiple Languages
N
raynardj
10.84k
11
Bioformer 8L
Apache-2.0
A lightweight BERT model specifically designed for biomedical text mining, running 3 times faster than BERT-base while achieving comparable or superior performance to BioBERT/PubMedBERT
Large Language Model Transformers English
B
bioformers
164
7
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase